Domain Adaptation In Reinforcement Learning Via Latent Unified State Representation

نویسندگان

چکیده

Despite the recent success of deep reinforcement learning (RL), domain adaptation remains an open problem. Although generalization ability RL agents is critical for real-world applicability Deep RL, zero-shot policy transfer still a challenging problem since even minor visual changes could make trained agent completely fail in new task. To address this issue, we propose two-stage that first learns latent unified state representation (LUSR) which consistent across multiple domains stage, and then do training one source based on LUSR second stage. The cross-domain consistency allows acquired from to generalize other target without extra training. We demonstrate our approach variants CarRacing games with customized manipulations, verify it CARLA, autonomous driving simulator more complex realistic observations. Our results show can achieve state-of-the-art performance related tasks outperforms prior approaches latent-representation image-to-image translation.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Deep Unsupervised Domain Adaptation for Image Classification via Low Rank Representation Learning

Domain adaptation is a powerful technique given a wide amount of labeled data from similar attributes in different domains. In real-world applications, there is a huge number of data but almost more of them are unlabeled. It is effective in image classification where it is expensive and time-consuming to obtain adequate label data. We propose a novel method named DALRRL, which consists of deep ...

متن کامل

Continuous-Domain Reinforcement Learning Using a Learned Qualitative State Representation

We present a method that allows an agent to learn a qualitative state representation that can be applied to reinforcement learning. By exploring the environment the agent is able to learn an abstraction that consists of landmarks that break the space into qualitative regions, and rules that predict changes in qualitative state. For each predictive rule the agent learns a context consisting of q...

متن کامل

Selecting the State-Representation in Reinforcement Learning

The problem of selecting the right state-representation in a reinforcement learning problem is considered. Several models (functions mapping past observations to a finite set) of the observations are given, and it is known that for at least one of these models the resulting state dynamics are indeed Markovian. Without knowing neither which of the models is the correct one, nor what are the prob...

متن کامل

Reinforcement learning with via-point representation

In this paper, we propose a new learning framework for motor control. This framework consists of two components: reinforcement learning and via-point representation. In the field of motor control, conventional reinforcement learning has been used to acquire control sequences such as cart-pole or stand-up robot control. Recently, researchers have become interested in hierarchical architecture, s...

متن کامل

Exploring Representation-Learning Approaches to Domain Adaptation

Most supervised language processing systems show a significant drop-off in performance when they are tested on text that comes from a domain significantly different from the domain of the training data. Sequence labeling systems like partof-speech taggers are typically trained on newswire text, and in tests their error rate on, for example, biomedical data can triple, or worse. We investigate t...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2021

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v35i12.17251